Goto

Collaborating Authors

 video capsule endoscopy


CASCRNet: An Atrous Spatial Pyramid Pooling and Shared Channel Residual based Network for Capsule Endoscopy

Srinanda, K V, Prabhu, M Manvith, Lal, Shyam

arXiv.org Artificial Intelligence

This manuscript summarizes work on the Capsule Vision Challenge 2024 by MISAHUB. To address the multi-class disease classification task, which is challenging due to the complexity and imbalance in the Capsule Vision challenge dataset, this paper proposes CASCRNet (Capsule endoscopy-Aspp-SCR-Network), a parameter-efficient and novel model that uses Shared Channel Residual (SCR) blocks and Atrous Spatial Pyramid Pooling (ASPP) blocks. Further, the performance of the proposed model is compared with other well-known approaches. The experimental results yield that proposed model provides better disease classification results. The proposed model was successful in classifying diseases with an F1 Score of 78.5% and a Mean AUC of 98.3%, which is promising given its compact architecture.


Optimizing Gastrointestinal Diagnostics: A CNN-Based Model for VCE Image Classification

Ahlawat, Vaneeta, Sharma, Rohit, Urush, null

arXiv.org Artificial Intelligence

In recent years, the diagnosis of gastrointestinal (GI) diseases has advanced greatly with the advent of high-tech video capsule endoscopy (VCE) technology, which allows for non-invasive observation of the digestive system. The MisaHub Capsule Vision Challenge encourages the development of vendor-independent artificial intelligence models that can autonomously classify GI anomalies from VCE images. This paper presents CNN architecture designed specifically for multiclass classification of ten gut pathologies, including angioectasia, bleeding, erosion, erythema, foreign bodies, lymphangiectasia, polyps, ulcers, and worms as well as their normal state.


Capsule Vision Challenge 2024: Multi-Class Abnormality Classification for Video Capsule Endoscopy

Bansal, Aakarsh, Singla, Bhuvanesh, Wankhade, Raajan Rajesh, Patil, Nagamma

arXiv.org Artificial Intelligence

This study presents an approach to developing a model for classifying abnormalities in video capsule endoscopy (VCE) frames. Given the challenges of data imbalance, we implemented a tiered augmentation strategy using the albumentations library to enhance minority class representation. Additionally, we addressed learning complexities by progressively structuring training tasks, allowing the model to differentiate between normal and abnormal cases and then gradually adding more specific classes based on data availability. Our pipeline, developed in PyTorch, employs a flexible architecture enabling seamless adjustments to classification complexity. We tested our approach using ResNet50 and a custom ViT-CNN hybrid model, with training conducted on the Kaggle platform. This work demonstrates a scalable approach to abnormality classification in VCE.


Integrating Deep Feature Extraction and Hybrid ResNet-DenseNet Model for Multi-Class Abnormality Detection in Endoscopic Images

Sagar, Aman, Mehta, Preeti, Shrivastva, Monika, Kumari, Suchi

arXiv.org Artificial Intelligence

Gastrointestinal (GI) and liver diseases have become increasingly prevalent across the globe, largely due to factors such as industrialization, dietary shifts, and the widespread use of antibiotics. These diseases pose significant diagnostic and treatment challenges, emphasizing the need for advanced medical technologies. Video Capsule Endoscopy (VCE) is a key non-invasive tool for examining the GI tract, especially in diagnosing conditions related to the small intestine, such as Crohn's disease, Celiac disease, and GI cancer. Unlike traditional endoscopy, VCE involves a small, pill-sized camera that travels through the digestive tract, capturing detailed images without sedation or invasive procedures. This method offers a comprehensive view of areas that are difficult to reach using conventional endoscopy. Despite its advantages, VCE faces challenges in practical application. A typical VCE procedure can generate between 57,000 to 1,000,000 images for 6-8 hours, which gastroenterologists must review.


The intersection of video capsule endoscopy and artificial intelligence: addressing unique challenges using machine learning

Guleria, Shan, Schwartz, Benjamin, Sharma, Yash, Fernandes, Philip, Jablonski, James, Adewole, Sodiq, Srivastava, Sanjana, Rhoads, Fisher, Porter, Michael, Yeghyayan, Michelle, Hyatt, Dylan, Copland, Andrew, Ehsan, Lubaina, Brown, Donald, Syed, Sana

arXiv.org Artificial Intelligence

Introduction: Technical burdens and time-intensive review processes limit the practical utility of video capsule endoscopy (VCE). Artificial intelligence (AI) is poised to address these limitations, but the intersection of AI and VCE reveals challenges that must first be overcome. We identified five challenges to address. Challenge #1: VCE data are stochastic and contains significant artifact. Challenge #2: VCE interpretation is cost-intensive. Challenge #3: VCE data are inherently imbalanced. Challenge #4: Existing VCE AIMLT are computationally cumbersome. Challenge #5: Clinicians are hesitant to accept AIMLT that cannot explain their process. Methods: An anatomic landmark detection model was used to test the application of convolutional neural networks (CNNs) to the task of classifying VCE data. We also created a tool that assists in expert annotation of VCE data. We then created more elaborate models using different approaches including a multi-frame approach, a CNN based on graph representation, and a few-shot approach based on meta-learning. Results: When used on full-length VCE footage, CNNs accurately identified anatomic landmarks (99.1%), with gradient weighted-class activation mapping showing the parts of each frame that the CNN used to make its decision. The graph CNN with weakly supervised learning (accuracy 89.9%, sensitivity of 91.1%), the few-shot model (accuracy 90.8%, precision 91.4%, sensitivity 90.9%), and the multi-frame model (accuracy 97.5%, precision 91.5%, sensitivity 94.8%) performed well. Discussion: Each of these five challenges is addressed, in part, by one of our AI-based models. Our goal of producing high performance using lightweight models that aim to improve clinician confidence was achieved.


Role of Artificial Intelligence in Video Capsule Endoscopy

#artificialintelligence

Capsule endoscopy (CE) has been increasingly utilised in recent years as a minimally invasive tool to investigate the whole gastrointestinal (GI) tract and a range of capsules are currently available for evaluation of upper GI, small bowel, and lower GI pathology. Although CE is undoubtedly an invaluable test for the investigation of small bowel pathology, it presents considerable challenges and limitations, such as long and laborious reading times, risk of missing lesions, lack of bowel cleansing score and lack of locomotion. Artificial intelligence (AI) seems to be a promising tool that may help improve the performance metrics of CE, and consequently translate to better patient care. In the last decade, significant progress has been made to apply AI in the field of endoscopy, including CE. Although it is certain that AI will find soon its place in day-to-day endoscopy clinical practice, there are still some open questions and barriers limiting its widespread application. In this review, we provide some general information about AI, and outline recent advances in AI and CE, issues around implementation of AI in medical practice and potential future applications of AI-aided CE.